ModernBERT is a modern bidirectional encoder-only Transformer model (BERT-style), pre-trained on 2 trillion tokens of English and code data, with a native context length of up to 8,192 tokens.
Large Language Model
Transformers Supports Multiple Languages